An efficient parallel block coordinate descent algorithm for large-scale precision matrix estimation using graphics processing units

نویسندگان

چکیده

Large-scale sparse precision matrix estimation has attracted wide interest from the statistics community. The convex partial correlation selection method (CONCORD) developed by Khare et al. (J R Stat Soc Ser B (Stat Methodol) 77(4):803–825, 2015) recently been credited with some theoretical properties for estimating matrices. CONCORD obtains its solution a coordinate descent algorithm (CONCORD-CD) based on convexity of objective function. However, since coordinate-wise update in CONCORD-CD is inherently serial, scale-up nontrivial. In this paper, we propose novel parallelization CONCORD-CD, namely, CONCORD-PCD. CONCORD-PCD partitions off-diagonal elements into several groups and updates each group simultaneously without harming computational convergence CONCORD-CD. We guarantee employing notion edge coloring graph theory. Specifically, establish nontrivial correspondence between scheduling edges complete graph. It turns out that simultanoeusly which associated are colorable same color. As result, number steps required updating reduces $$p(p-1)/2$$ to $$p-1$$ (for even p) or p odd p), where denotes variables. prove such irreducible addition, tailored single-instruction multiple-data (SIMD) parallelism. A numerical study shows SIMD-parallelized PCD implemented graphics processing units boosts multiple times. available package pcdconcord.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Large Scale Kernel Learning using Block Coordinate Descent

We demonstrate that distributed block coordinate descent can quickly solve kernel regression and classification problems with millions of data points. Armed with this capability, we conduct a thorough comparison between the full kernel, the Nyström method, and random features on three large classification tasks from various domains. Our results suggest that the Nyström method generally achieves...

متن کامل

Greedy Block Coordinate Descent for Large Scale Gaussian Process Regression

We propose a variable decomposition algorithm– greedy block coordinate descent (GBCD)–in order to make dense Gaussian process regression practical for large scale problems. GBCD breaks a large scale optimization into a series of small sub-problems. The challenge in variable decomposition algorithms is the identification of a subproblem (the active set of variables) that yields the largest impro...

متن کامل

An Asynchronous Parallel Stochastic Coordinate Descent Algorithm

We describe an asynchronous parallel stochastic coordinate descent algorithm for minimizing smooth unconstrained or separably constrained functions. The method achieves a linear convergence rate on functions that satisfy an essential strong convexity property and a sublinear rate (1/K) on general convex functions. Near-linear speedup on a multicore system can be expected if the number of proces...

متن کامل

Stochastic Parallel Block Coordinate Descent for Large-Scale Saddle Point Problems

We consider convex-concave saddle point problems with a separable structure and non-strongly convex functions. We propose an efficient stochastic block coordinate descent method using adaptive primal-dual updates, which enables flexible parallel optimization for large-scale problems. Our method shares the efficiency and flexibility of block coordinate descent methods with the simplicity of prim...

متن کامل

A Block-Coordinate Descent Approach for Large-scale Sparse Inverse Covariance Estimation

The sparse inverse covariance estimation problem arises in many statistical applications in machine learning and signal processing. In this problem, the inverse of a covariance matrix of a multivariate normal distribution is estimated, assuming that it is sparse. An `1 regularized log-determinant optimization problem is typically solved to approximate such matrices. Because of memory limitation...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Computational Statistics

سال: 2021

ISSN: ['0943-4062', '1613-9658']

DOI: https://doi.org/10.1007/s00180-021-01127-x